Application of implicit knowledge: deterministic or probabilistic?

نویسندگان

  • Zoltán Dienes
  • Andreas Kurz
  • Regina Bernhaupt
  • Josef Perner
چکیده

This paper distinguishes two models specifying the application of implicit knowledge. According to one model, originally suggested by Reber (1967), subjects either apply sufficient knowledge to always produce a correct response or else they guess randomly (High Threshold Theory; subjects only apply knowledge when there is sufficient knowledge to exceed a threshold ensuring a correct response); according to the other model, suggested by Dienes (1992), subjects respond with a certain probability towards each item, where the probability is determined by the match between the items structure and the induced constraints about the structure (Probability Matching Theory; subjects match their probability of responding against their personal probability that the item belongs to a certain category). One parameter versions of both models were specified and then tested against the data generated from three artificial grammar learning experiments. Neither theory could account for all features of the data, and extensions of the theories are suggested. Dienes and Berry (1997) argued that there is widespread agreement about the existence of an important learning mechanism, pervasive in its effects, and producing knowledge by unconscious associative processes, knowledge about which the subject does not directly have metaknowledge. Let us call the mechanism implicit learning, and the induced knowledge implicit knowledge. This paper will address the question of how implicit knowledge is applied. Imagine an animal learning about which stimuli constitute food. Maybe its mother brings back different plants or animals for it to eat, and stops it from ingesting other plants or animals, which may be dangerous or poisonous. The details about what features go together to constitute something that is edible may be more or less complex; in any case, the animal eventually learns, let us say perfectly, to seek out only the edible substances in its environment. But before learning reaches this point, how should the animal behave towards stimuli about which it has only imperfect information and when the mother is not around to provide corrective feedback? One strategy would be to only ingest stimuli about which the animal’s induced knowledge unambiguously indicates are edible; this would prevent unfortunate fatalities. On this scenario, implicit knowledge may have evolved in general to direct action towards a stimulus only when the implicit knowledge relevant to the stimulus unambiguously indicates the category to which the stimulus belongs. Partial information may not even be made available for controlling behaviour in case the animal uses the knowledge to sample poisonous substances. Thus, if the animal were forced to respond to some stimuli about which it had imperfect information, the animal would be reduced to random responding. We will call this the High Threshold Theory (HTT) metaphorically, the knowledge lies fallow until it exceeds a high threshold. Now imagine an animal learning about which locations are likely to contain food. This scene differs from the last in that a wrong decision is not so likely to have catastrophic consequences. If the animal’s knowledge merely suggests that a location contains food, it still may be worth sampling in preference to another. There are two different states of affairs to be distinguished in this case: First, different locations may have objectively different probabilities of containing food. The animal may come to have perfect knowledge of the different probabilities. It may seem rational for the animal to always choose the location with the highest probability, but this is not so if the probability structure of the environment is subject to fluctuations. Then it is beneficial to occasionally sample even low probability locations because it allows an assessment of whether the probability remains low. Also, because other animals will in general be competing for the food, and there will be more competitiors at higher probability locations, it therefore pays to forage sometimes at low probability locations. In fact, animals do forage in different locations according to the probability of finding food there (Stephens & Krebs, 1986) that is, they probability match. Similarly, in predicting probabilistic events, people like other animals are also known to probability match rather than respond deterministically. Reber and Millward (1968) asked people to passively observe binary events at a rate of two per second for two or three minutes. When people were then asked to predict successive events they probability matched to a high degree of accuracy (to the second decimal place; see Reber, 1989, for review). Second, different locations may have imperfectly known probabilities of containing food. This is the case in which we are interested; how does the animal apply its imperfect knowledge? The knowledge may be imperfect because the locations have not yet been sampled frequently, or because the location is new and the animal does not know how to classify it. Consider an animal faced with choosing one of two locations. In the absence of certain knowledge about the objective probability structure, one strategy would be to search each location according to the estimated probability. For example, in the absence of any information, the animal would search each location with equal probability, and adjust the search probabilities as information was collected. Or if features of the locations suggested different probabilities, search could start at those estimated probabilities. The true probabilities could be honed in on more quickly if the animal starts near their actual values. We will call the hypothesis that the animal responds to stimuli with a probability that matches the stimuli’s expected probabilties Probability Matching Theory (PMT). If we allow probabilities to apply to singular events, then animals may respond probabilistically to stimuli for which there are no objective probabilities other than 0 or 1. For example, in classifying an edible stimulus as either being palatable or unpalatable, the animal may ingest it with a probability determined by the likelihood that the stimulus is palatable. In this case, the probability of the animal ingesting a stimulus may be interpreted as the animal’s ‘personal probability’ that the stimulus is edible. We have outlined two hypotheses about the way in which implicit knowledge could be applied: HTT and PMT. It may be that implicit knowledge is applied in different ways according to context; for example, according to the cost of a wrong choice, as in the examples of ingesting stimuli and searching locations given above. Or it may be that evolution has for economy produced a mechanism with a single principle of application. We will investigate these two hypotheses HTT and PMT in one context, that of people implicitly learning artificial grammars. In that field, both hypotheses have been suggested as accounting for human performance. In a typical artificial grammar learning experiment, subjects first memorize grammatical strings of letters generated by a finite-state grammar. Then, they are informed of the existence of the complex set of rules that constrains letter order (but not what they are), and are asked to classify grammatical and nongrammatical strings. In an initial study, Reber (1967) found that the more strings subjects had attempted to memorize, the easier it was to memorize novel grammatical strings, indicating that they had learned to utilize the structure of the grammar. Subjects could also classify novel strings significantly above chance (69%, where chance is 50%). Reber (e.g. 1989) argued that the knowledge was implicit because subjects could not adequately describe how they classified strings (see Dienes & Perner, 1996, and Dienes & Berry, 1997, for further arguments that the knowledge should be seen as implicit). Reber (1967, 1989) argued for a theory of implicit learning which combined HTT with the Gibsonian notion of veridical information pick-up. Specifically, Reber argued that implicit learning results in underlying representations that mirror the objective structures in the environment. In the case of artificial grammar learning, he argued that the status of a given test item would either be known (and would thus always be classified correctly); or the status was not known, and the subject guessed randomly. On this assumption, any knowledge the subject applied was always perfectly veridical; incorrect responses could only be based on failing to apply knowledge rather than on applying incorrect knowledge. Reber used these assumptions by testing each subject twice on each string without feedback. If the probability of a given subject using perfect knowledge of the ith string is ki then the expected proportion of strings classified correctly twice, once, or not at all by that subject is determined by the following equations: the proportion of strings classified twice correctly = p(CC) = k + (1-k)*0.5 the proportion of strings classified correctly just once = p(CE) + p(EC) = (1-k)* 0.5 the proportion of strings never classified correctly = p(EE) = (1-k)* 0.5 where k is the average of the kis. Under this model, the values of p(CE), p(EC) and p(EE) averaged across subjects should be statistically identical and lower than p(CC). If p(EE) is greater than p(CE) or p(EC), then this is evidence that subjects have induced rules that are not accurate reflections of the grammar; the incorrect rules lead to consistent misclassifications. Reber (1989) reviewed 8 studies in which subjects in the learning phase were exposed to grammatical stimuli presented in a random order so that the grammatical rules would not be salient, and in which in the test phase subjects were tested twice. When subjects were asked to search for rules in the learning phase, p(EE) was on average .22 and the average of p(CE) and P(EC) was .13. That is, when subjects were asked to learn explicitly they induced substantial amounts of nonrepresentative rules. When subjects were asked to memorize strings in the learning phase, p(EE) was on average .16 and the average of p(CE) and p(EC) was .12. That is, when subjects were encouraged to learn implicitly, subjects only applied knowledge when it was largely (even if not entirely) representative, consistent with Reber’s claim that knowledge only applied when it unambiguously indicated a correct classification. In contrast to Reber, Dienes (1992; Dienes & Perner, 1996) suggested a connectionist model of artificial grammar learning that applied its knowledge according to PMT. The model became sensitive to the actual statistical structure of the environment, and to this degree embodied Reber’s notion of representational realism. However, the network was a two-layer connectionist network and so could only become sensitive to first order statistical structure, and not higher orders; in this way, it’s knowledge did not entirely mirror that of the environment. During the learning phase the network was trained to predict each letter from every other letter in the learning string. In the test phase, it attempted the same prediction. The activation of an ouput unit was determined by the correlation between the actual and predicted letters in the test string. The output activation directly determined the probability that the network would give a ‘grammatical’ response. Thus, the more each letter in the test string could be accurately predicted from the other letters, the greater was the probability of the network saying ‘grammatical’. Dienes (1992) showed that for a set of stimuli that had been used in the past in human experiments by both Dulany, Carlson, and Dewey (1984) and Dienes, Broadbent, and Berry (1991), the proportion of subjects classifying each test string correctly was a reliable feature of the data across different experiments. The proportions for different strings varied between 37% and 82% with a good spread inbetween. Reber’s model would interpret this variation in proportions as variation in the proportion of subjects who knew each string (with the values below 50% being due to the occasional nonrepresentative knowledge induced by subjects). Dienes interpreted the variation not as variation across subjects but as a variation across items in the probability of saying ‘grammatical’ (it was assumed that this probability was constant across subjects). In other words, subjects could be ‘probability matching’ (in quotes because there are no objective probabilities to match each test string either is or is not grammatical). The model incorporated the assumption of ‘probability matching’ in that the activation of the output unit codes the degree that the test string satisfies the internal constraints induced by the grammar, and hence a sort of personal probability that the string is grammatical. Dienes (1992) showed that the connectionist network could successfully predict the probability of each test string being called grammatical (i.e. the proportion of subjects who called it grammatical), the overall classification performance of subjects, and the p(CC), p(CE), and p(EE) values averaged over subjects and strings. That is, Dienes showed that a model incorporating the assumptions of PMT could account for the same data that Reber (1989) had argued supported his HTT. The closeness of p(CE) and p(EE) could only be produced in the PMT model by having some proportion of items with a probability correct, p, well below .50. If p > .50, then p(CE) > p(EE) since p(1-p) > (1-p). In summary, the data published so far do not discriminate the two theories of application of implicit knowledge, PMT and HTT. The aim of this paper is to directly test the two theories by testing each subject three times on a set of test strings and looking at the distribution of correct responses for each string separately. If people are tested three times on each string instead of twice, the Reber model makes the following predictions, where X is the proportion of people who know the string: the proportion of people classifying an item three times correctly = p(3Cs) = X + (1-X)*0.5 the proportion of people making two correct classifications and one error = p(2Cs, 1E) = (1-X)*3*0.5, the proportion of people making two errors and one correct classification = p(1C, 2Es) = (1-X)*3*0.5 and the proportion of people making three errors = p(3Es) = (1-X)* 0.5. According to the PMT model, the corresponding equations are p(3Cs) = p p(2Cs, 1E) = 3*p*(1-p) p(1C, 2Es) = 3*p*(1-p) 2 p(3Es) = (1-p) 3 where p is the probability of responding correctly to the string. Both models contain just one parameter which can be estimated from the overall proportion, P, of correct classifications for each item. For the Reber model, X can be estimated from the relationship X + (1-X)*0.5 = P (thus, X = 2*P 1), and for the PMT model p is estimated directly by P. In some cases, the models make qualitatively similar predictions. For example, the Reber model predicts that p(3Cs) should be greater than p(2Cs, 1E) given that P > .70. Similarly, the PMT model predicts that p(3Cs) should be greater than p(2Cs, 1E) given that P > .75. Further, both models predict that for Ps close to .50, the distribution of responses should approximate a binomial with p=0.50. The models also make qualitatively different predictions. For the Reber model, regardless of P, p(2Cs,1E) should always be equal to p(1C,2Es). For the PMT model, these probabilities will differ except in the special case that p = 0.5. If p > 0.5 then p(2Cs,1E) will be greater than p(1C,2Es). Also for the Reber model, regardless of P, p(1C,2Es) should always be equal to 3*p(3Es). For the PMT model, p(1C,2Es) will be greater than 3*p(3Es) whenever p > 0.5. The models can also be fitted quantitatively to the actual distributions obtained. These qualitative and quantitative predictions are explored in three experiments.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Implicational Scaling of Reading Comprehension Construct: Is it Deterministic or Probabilistic?

In English as a Second Language Teaching and Testing situations, it is common to infer about learners’ reading ability based on his or her total score on a reading test. This assumes the unidimensional and reproducible nature of reading items. However, few researches have been conducted to probe the issue through psychometric analyses. In the present study, the IELTS exemplar module C (1994) wa...

متن کامل

A Fire Ignition Model and Its Application for Estimating Loss due to Damage of the Urban Gas Network in an Earthquake

Damage of the urban gas network due to an earthquake can cause much loss including fire-induced loss to infrastructure and loss due to interruption of gas service and repairing or replacing of network elements. In this paper, a new fire ignition model is proposed and applied to a conventional semi-probabilistic model for estimating various losses due to damage of an urban gas network in an eart...

متن کامل

Probabilistic design: The future of rock engineering

A brief background to the development of the rock engineering design process is given, showing that since the development of the science of mathematics, deterministic methods have been used to perform various calculations. The variability of rock properties and support characteristics have always been known. However, they were not explicitly used in design but compensated for by the use of a sa...

متن کامل

Study of Explicit Knowledge Effects on Implicit Motor Learning in Older Adults

Objective: The motor skill performance and learning in older adults are of a great importance. The primary purpose of this study was to determine whether older adult’s explicit knowledge effect on implicit learning and motor performance and which are typical for rehabilitation and skills acquired in older adults. Materials & Methods: In this comparative study a serial reaction time task by u...

متن کامل

Probabilistic Recursion Theory and Implicit Computational Complexity

In this thesis we provide a characterization of probabilistic computation in itself, from a recursion-theoretical perspective, without reducing it to deterministic computation. More specifically, we show that probabilistic computable functions, i.e., those functions which are computed by Probabilistic Turing Machines (PTM), can be characterized by a natural generalization of Kleene’s partial re...

متن کامل

Evaluating the Effectiveness of Explicit and Implicit Form-Focused Instruction on Explicit and Implicit Knowledge of EFL Learners

Although explicit and implicit knowledge of language learners are essential to theoretical and pedagogical debates in second language acquisition (SLA), little research has addressed the effects of instructional interventions on the two knowledge types (R. Ellis, 2005).This study examined the relative effectiveness of explicit and implicit types of form-focused instruction (FFI) on the acquisit...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997